summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorH Lohaus <hlohaus@users.noreply.github.com>2024-01-26 07:54:13 +0100
committerGitHub <noreply@github.com>2024-01-26 07:54:13 +0100
commitfeb83c168b0a57ecd8c511aa654209c5f40da30e (patch)
tree84ef9a576064b7480e339426d9966b17a3102cad
parentMerge pull request #1510 from hlohaus/sort (diff)
downloadgpt4free-0.2.0.6.tar
gpt4free-0.2.0.6.tar.gz
gpt4free-0.2.0.6.tar.bz2
gpt4free-0.2.0.6.tar.lz
gpt4free-0.2.0.6.tar.xz
gpt4free-0.2.0.6.tar.zst
gpt4free-0.2.0.6.zip
-rw-r--r--.github/workflows/unittest.yml13
-rw-r--r--README.md12
-rw-r--r--etc/unittest/asyncio.py8
-rw-r--r--etc/unittest/backend.py12
-rw-r--r--g4f/Provider/Bing.py31
-rw-r--r--g4f/Provider/DeepInfra.py11
-rw-r--r--g4f/Provider/GptForLove.py11
-rw-r--r--g4f/Provider/HuggingChat.py2
-rw-r--r--g4f/Provider/PerplexityLabs.py8
-rw-r--r--g4f/Provider/Phind.py1
-rw-r--r--g4f/Provider/Vercel.py15
-rw-r--r--g4f/Provider/base_provider.py1
-rw-r--r--g4f/Provider/bing/conversation.py2
-rw-r--r--g4f/Provider/bing/create_images.py24
-rw-r--r--g4f/Provider/bing/upload_image.py51
-rw-r--r--g4f/Provider/deprecated/ChatgptDuo.py4
-rw-r--r--g4f/Provider/deprecated/GetGpt.py30
-rw-r--r--g4f/Provider/helper.py68
-rw-r--r--g4f/Provider/needs_auth/Bard.py12
-rw-r--r--g4f/Provider/needs_auth/OpenaiChat.py52
-rw-r--r--g4f/Provider/selenium/PerplexityAi.py12
-rw-r--r--g4f/base_provider.py2
-rw-r--r--g4f/cli.py2
-rw-r--r--g4f/errors.py8
-rw-r--r--g4f/gui/__init__.py10
-rw-r--r--g4f/gui/run.py8
-rw-r--r--g4f/gui/server/internet.py12
-rw-r--r--g4f/image.py60
-rw-r--r--g4f/requests.py190
-rw-r--r--g4f/typing.py6
-rw-r--r--g4f/version.py2
-rw-r--r--g4f/webdriver.py24
-rw-r--r--requirements-min.txt2
-rw-r--r--requirements.txt2
-rw-r--r--setup.py75
35 files changed, 485 insertions, 298 deletions
diff --git a/.github/workflows/unittest.yml b/.github/workflows/unittest.yml
index aadc7b86..90558f09 100644
--- a/.github/workflows/unittest.yml
+++ b/.github/workflows/unittest.yml
@@ -15,10 +15,19 @@ jobs:
runs-on: ubuntu-latest
steps:
- uses: actions/checkout@v4
- - name: Set up Python
+ - name: Set up Python 3.8
uses: actions/setup-python@v4
with:
- python-version: "3.x"
+ python-version: "3.8"
+ cache: 'pip'
+ - name: Install min requirements
+ run: pip install -r requirements-min.txt
+ - name: Run tests
+ run: python -m etc.unittest
+ - name: Set up Python 3.11
+ uses: actions/setup-python@v4
+ with:
+ python-version: "3.11"
cache: 'pip'
- name: Install requirements
run: pip install -r requirements.txt
diff --git a/README.md b/README.md
index a4365e09..da994d87 100644
--- a/README.md
+++ b/README.md
@@ -100,7 +100,7 @@ or set the api base in your client to: [http://localhost:1337/v1](http://localho
##### Install using pypi:
```
-pip install -U g4f
+pip install -U "g4f[all]"
```
##### or:
@@ -134,13 +134,19 @@ python3 -m venv venv
```
source venv/bin/activate
```
-5. Install the required Python packages from `requirements.txt`:
+5. Install minimum requirements:
+
+```
+pip install -r requirements-min.txt
+```
+
+6. Or install all used Python packages from `requirements.txt`:
```
pip install -r requirements.txt
```
-6. Create a `test.py` file in the root folder and start using the repo, further Instructions are below
+7. Create a `test.py` file in the root folder and start using the repo, further Instructions are below
```py
import g4f
diff --git a/etc/unittest/asyncio.py b/etc/unittest/asyncio.py
index 74e29986..a31ce211 100644
--- a/etc/unittest/asyncio.py
+++ b/etc/unittest/asyncio.py
@@ -1,6 +1,10 @@
from .include import DEFAULT_MESSAGES
import asyncio
-import nest_asyncio
+try:
+ import nest_asyncio
+ has_nest_asyncio = True
+except:
+ has_nest_asyncio = False
import unittest
import g4f
from g4f import ChatCompletion
@@ -39,6 +43,8 @@ class TestChatCompletionAsync(unittest.IsolatedAsyncioTestCase):
class TestChatCompletionNestAsync(unittest.IsolatedAsyncioTestCase):
def setUp(self) -> None:
+ if not has_nest_asyncio:
+ self.skipTest('"nest_asyncio" not installed')
nest_asyncio.apply()
async def test_create(self):
diff --git a/etc/unittest/backend.py b/etc/unittest/backend.py
index f5961e2d..3be83f84 100644
--- a/etc/unittest/backend.py
+++ b/etc/unittest/backend.py
@@ -3,11 +3,17 @@ import unittest
from unittest.mock import MagicMock
from .mocks import ProviderMock
import g4f
-from g4f.gui.server.backend import Backend_Api, get_error_message
+try:
+ from g4f.gui.server.backend import Backend_Api, get_error_message
+ has_requirements = True
+except:
+ has_requirements = False
class TestBackendApi(unittest.TestCase):
def setUp(self):
+ if not has_requirements:
+ self.skipTest('"flask" not installed')
self.app = MagicMock()
self.api = Backend_Api(self.app)
@@ -28,6 +34,10 @@ class TestBackendApi(unittest.TestCase):
class TestUtilityFunctions(unittest.TestCase):
+ def setUp(self):
+ if not has_requirements:
+ self.skipTest('"flask" not installed')
+
def test_get_error_message(self):
g4f.debug.last_provider = ProviderMock
exception = Exception("Message")
diff --git a/g4f/Provider/Bing.py b/g4f/Provider/Bing.py
index 11bb1414..32879fa6 100644
--- a/g4f/Provider/Bing.py
+++ b/g4f/Provider/Bing.py
@@ -9,7 +9,7 @@ from urllib import parse
from aiohttp import ClientSession, ClientTimeout, BaseConnector
from ..typing import AsyncResult, Messages, ImageType
-from ..image import ImageResponse
+from ..image import ImageResponse, ImageRequest
from .base_provider import AsyncGeneratorProvider
from .helper import get_connector
from .bing.upload_image import upload_image
@@ -154,6 +154,11 @@ class Defaults:
'SRCHHPGUSR' : f'HV={int(time.time())}',
}
+class ConversationStyleOptionSets():
+ CREATIVE = ["h3imaginative", "clgalileo", "gencontentv3"]
+ BALANCED = ["galileo"]
+ PRECISE = ["h3precise", "clgalileo"]
+
def format_message(msg: dict) -> str:
"""
Formats a message dictionary into a JSON string with a delimiter.
@@ -168,7 +173,7 @@ def create_message(
prompt: str,
tone: str,
context: str = None,
- image_response: ImageResponse = None,
+ image_request: ImageRequest = None,
web_search: bool = False,
gpt4_turbo: bool = False
) -> str:
@@ -179,7 +184,7 @@ def create_message(
:param prompt: The user's input prompt.
:param tone: The desired tone for the response.
:param context: Additional context for the prompt.
- :param image_response: The response if an image is involved.
+ :param image_request: The image request with the url.
:param web_search: Flag to enable web search.
:param gpt4_turbo: Flag to enable GPT-4 Turbo.
:return: A formatted string message for the Bing API.
@@ -187,11 +192,11 @@ def create_message(
options_sets = Defaults.optionsSets
# Append tone-specific options
if tone == Tones.creative:
- options_sets.append("h3imaginative")
+ options_sets.extend(ConversationStyleOptionSets.CREATIVE)
elif tone == Tones.precise:
- options_sets.append("h3precise")
+ options_sets.extend(ConversationStyleOptionSets.PRECISE)
elif tone == Tones.balanced:
- options_sets.append("galileo")
+ options_sets.extend(ConversationStyleOptionSets.BALANCED)
else:
options_sets.append("harmonyv3")
@@ -233,9 +238,9 @@ def create_message(
'type': 4
}
- if image_response and image_response.get('imageUrl') and image_response.get('originalImageUrl'):
- struct['arguments'][0]['message']['originalImageUrl'] = image_response.get('originalImageUrl')
- struct['arguments'][0]['message']['imageUrl'] = image_response.get('imageUrl')
+ if image_request and image_request.get('imageUrl') and image_request.get('originalImageUrl'):
+ struct['arguments'][0]['message']['originalImageUrl'] = image_request.get('originalImageUrl')
+ struct['arguments'][0]['message']['imageUrl'] = image_request.get('imageUrl')
struct['arguments'][0]['experienceType'] = None
struct['arguments'][0]['attachedFileInfo'] = {"fileName": None, "fileType": None}
@@ -282,9 +287,9 @@ async def stream_generate(
timeout=ClientTimeout(total=timeout), headers=headers, connector=connector
) as session:
conversation = await create_conversation(session)
- image_response = await upload_image(session, image, tone) if image else None
- if image_response:
- yield image_response
+ image_request = await upload_image(session, image, tone) if image else None
+ if image_request:
+ yield image_request
try:
async with session.ws_connect(
@@ -294,7 +299,7 @@ async def stream_generate(
) as wss:
await wss.send_str(format_message({'protocol': 'json', 'version': 1}))
await wss.receive(timeout=timeout)
- await wss.send_str(create_message(conversation, prompt, tone, context, image_response, web_search, gpt4_turbo))
+ await wss.send_str(create_message(conversation, prompt, tone, context, image_request, web_search, gpt4_turbo))
response_txt = ''
returned_text = ''
diff --git a/g4f/Provider/DeepInfra.py b/g4f/Provider/DeepInfra.py
index 2f34b679..09b9464e 100644
--- a/g4f/Provider/DeepInfra.py
+++ b/g4f/Provider/DeepInfra.py
@@ -13,11 +13,12 @@ class DeepInfra(AsyncGeneratorProvider, ProviderModelMixin):
supports_message_history = True
default_model = 'meta-llama/Llama-2-70b-chat-hf'
- @staticmethod
- def get_models():
- url = 'https://api.deepinfra.com/models/featured'
- models = requests.get(url).json()
- return [model['model_name'] for model in models]
+ @classmethod
+ def get_models(cls):
+ if not cls.models:
+ url = 'https://api.deepinfra.com/models/featured'
+ cls.models = requests.get(url).json()
+ return cls.models
@classmethod
async def create_async_generator(
diff --git a/g4f/Provider/GptForLove.py b/g4f/Provider/GptForLove.py
index 07e3406f..cc82da21 100644
--- a/g4f/Provider/GptForLove.py
+++ b/g4f/Provider/GptForLove.py
@@ -1,11 +1,18 @@
from __future__ import annotations
from aiohttp import ClientSession
-import execjs, os, json
+import os
+import json
+try:
+ import execjs
+ has_requirements = True
+except ImportError:
+ has_requirements = False
from ..typing import AsyncResult, Messages
from .base_provider import AsyncGeneratorProvider
from .helper import format_prompt
+from ..errors import MissingRequirementsError
class GptForLove(AsyncGeneratorProvider):
url = "https://ai18.gptforlove.com"
@@ -20,6 +27,8 @@ class GptForLove(AsyncGeneratorProvider):
proxy: str = None,
**kwargs
) -> AsyncResult:
+ if not has_requirements:
+ raise MissingRequirementsError('Install "PyExecJS" package')
if not model:
model = "gpt-3.5-turbo"
headers = {
diff --git a/g4f/Provider/HuggingChat.py b/g4f/Provider/HuggingChat.py
index 79e4ae38..9aa93878 100644
--- a/g4f/Provider/HuggingChat.py
+++ b/g4f/Provider/HuggingChat.py
@@ -39,7 +39,7 @@ class HuggingChat(AsyncGeneratorProvider, ProviderModelMixin):
**kwargs
) -> AsyncResult:
if not cookies:
- cookies = get_cookies(".huggingface.co")
+ cookies = get_cookies(".huggingface.co", False)
headers = {
'User-Agent': 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_15_7) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/111.0.0.0 Safari/537.36',
diff --git a/g4f/Provider/PerplexityLabs.py b/g4f/Provider/PerplexityLabs.py
index 5002b39f..a7b98f7c 100644
--- a/g4f/Provider/PerplexityLabs.py
+++ b/g4f/Provider/PerplexityLabs.py
@@ -14,12 +14,12 @@ WS_URL = "wss://labs-api.perplexity.ai/socket.io/"
class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
url = "https://labs.perplexity.ai"
working = True
+ default_model = 'pplx-70b-online'
models = [
'pplx-7b-online', 'pplx-70b-online', 'pplx-7b-chat', 'pplx-70b-chat', 'mistral-7b-instruct',
'codellama-34b-instruct', 'llama-2-70b-chat', 'llava-7b-chat', 'mixtral-8x7b-instruct',
'mistral-medium', 'related'
]
- default_model = 'pplx-70b-online'
model_aliases = {
"mistralai/Mistral-7B-Instruct-v0.1": "mistral-7b-instruct",
"meta-llama/Llama-2-70b-chat-hf": "llama-2-70b-chat",
@@ -52,8 +52,7 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
async with ClientSession(headers=headers, connector=get_connector(connector, proxy)) as session:
t = format(random.getrandbits(32), '08x')
async with session.get(
- f"{API_URL}?EIO=4&transport=polling&t={t}",
- proxy=proxy
+ f"{API_URL}?EIO=4&transport=polling&t={t}"
) as response:
text = await response.text()
@@ -61,8 +60,7 @@ class PerplexityLabs(AsyncGeneratorProvider, ProviderModelMixin):
post_data = '40{"jwt":"anonymous-ask-user"}'
async with session.post(
f'{API_URL}?EIO=4&transport=polling&t={t}&sid={sid}',
- data=post_data,
- proxy=proxy
+ data=post_data
) as response:
assert await response.text() == 'OK'
diff --git a/g4f/Provider/Phind.py b/g4f/Provider/Phind.py
index dfc8f992..dbf1e7ae 100644
--- a/g4f/Provider/Phind.py
+++ b/g4f/Provider/Phind.py
@@ -9,7 +9,6 @@ from ..requests import StreamSession
class Phind(AsyncGeneratorProvider):
url = "https://www.phind.com"
working = True
- supports_gpt_4 = True
supports_stream = True
supports_message_history = True
diff --git a/g4f/Provider/Vercel.py b/g4f/Provider/Vercel.py
index 466ea3de..8d2137bf 100644
--- a/g4f/Provider/Vercel.py
+++ b/g4f/Provider/Vercel.py
@@ -1,10 +1,16 @@
from __future__ import annotations
-import json, base64, requests, execjs, random, uuid
+import json, base64, requests, random, uuid
+
+try:
+ import execjs
+ has_requirements = True
+except ImportError:
+ has_requirements = False
from ..typing import Messages, TypedDict, CreateResult, Any
from .base_provider import AbstractProvider
-from ..debug import logging
+from ..errors import MissingRequirementsError
class Vercel(AbstractProvider):
url = 'https://sdk.vercel.ai'
@@ -21,10 +27,11 @@ class Vercel(AbstractProvider):
proxy: str = None,
**kwargs
) -> CreateResult:
-
+ if not has_requirements:
+ raise MissingRequirementsError('Install "PyExecJS" package')
+
if not model:
model = "gpt-3.5-turbo"
-
elif model not in model_info:
raise ValueError(f"Vercel does not support {model}")
diff --git a/g4f/Provider/base_provider.py b/g4f/Provider/base_provider.py
index e1dcd24d..b173db4e 100644
--- a/g4f/Provider/base_provider.py
+++ b/g4f/Provider/base_provider.py
@@ -1,4 +1,5 @@
from __future__ import annotations
+
import sys
import asyncio
from asyncio import AbstractEventLoop
diff --git a/g4f/Provider/bing/conversation.py b/g4f/Provider/bing/conversation.py
index 36ada3b0..388bdd6b 100644
--- a/g4f/Provider/bing/conversation.py
+++ b/g4f/Provider/bing/conversation.py
@@ -1,3 +1,5 @@
+from __future__ import annotations
+
from aiohttp import ClientSession
class Conversation:
diff --git a/g4f/Provider/bing/create_images.py b/g4f/Provider/bing/create_images.py
index af39ef1e..4fa85929 100644
--- a/g4f/Provider/bing/create_images.py
+++ b/g4f/Provider/bing/create_images.py
@@ -2,21 +2,28 @@
This module provides functionalities for creating and managing images using Bing's service.
It includes functions for user login, session creation, image creation, and processing.
"""
+from __future__ import annotations
import asyncio
import time
import json
import os
from aiohttp import ClientSession, BaseConnector
-from bs4 import BeautifulSoup
from urllib.parse import quote
from typing import Generator, List, Dict
+try:
+ from bs4 import BeautifulSoup
+ has_requirements = True
+except ImportError:
+ has_requirements = False
+
from ..create_images import CreateImagesProvider
from ..helper import get_cookies, get_connector
from ...webdriver import WebDriver, get_driver_cookies, get_browser
from ...base_provider import ProviderType
from ...image import ImageResponse
+from ...errors import MissingRequirementsError, MissingAccessToken
BING_URL = "https://www.bing.com"
TIMEOUT_LOGIN = 1200
@@ -97,6 +104,8 @@ async def create_images(session: ClientSession, prompt: str, proxy: str = None,
Raises:
RuntimeError: If image creation fails or times out.
"""
+ if not has_requirements:
+ raise MissingRequirementsError('Install "beautifulsoup4" package')
url_encoded_prompt = quote(prompt)
payload = f"q={url_encoded_prompt}&rt=4&FORM=GENCRE"
url = f"{BING_URL}/images/create?q={url_encoded_prompt}&rt=4&FORM=GENCRE"
@@ -193,7 +202,11 @@ class CreateImagesBing:
Yields:
Generator[str, None, None]: The final output as markdown formatted string with images.
"""
- cookies = self.cookies or get_cookies(".bing.com")
+ try:
+ cookies = self.cookies or get_cookies(".bing.com")
+ except MissingRequirementsError as e:
+ raise MissingAccessToken(f'Missing "_U" cookie. {e}')
+
if "_U" not in cookies:
login_url = os.environ.get("G4F_LOGIN_URL")
if login_url:
@@ -211,9 +224,12 @@ class CreateImagesBing:
Returns:
str: Markdown formatted string with images.
"""
- cookies = self.cookies or get_cookies(".bing.com")
+ try:
+ cookies = self.cookies or get_cookies(".bing.com")
+ except MissingRequirementsError as e:
+ raise MissingAccessToken(f'Missing "_U" cookie. {e}')
if "_U" not in cookies:
- raise RuntimeError('"_U" cookie is missing')
+ raise MissingAccessToken('Missing "_U" cookie')
proxy = os.environ.get("G4F_PROXY")
async with create_session(cookies, proxy) as session:
images = await create_images(session, prompt, self.proxy)
diff --git a/g4f/Provider/bing/upload_image.py b/g4f/Provider/bing/upload_image.py
index bb5687a8..f9e11561 100644
--- a/g4f/Provider/bing/upload_image.py
+++ b/g4f/Provider/bing/upload_image.py
@@ -1,17 +1,14 @@
"""
Module to handle image uploading and processing for Bing AI integrations.
"""
-
from __future__ import annotations
-import string
-import random
+
import json
import math
-from aiohttp import ClientSession
-from PIL import Image
+from aiohttp import ClientSession, FormData
from ...typing import ImageType, Tuple
-from ...image import to_image, process_image, to_base64, ImageResponse
+from ...image import to_image, process_image, to_base64_jpg, ImageRequest, Image
IMAGE_CONFIG = {
"maxImagePixels": 360000,
@@ -24,7 +21,7 @@ async def upload_image(
image_data: ImageType,
tone: str,
proxy: str = None
-) -> ImageResponse:
+) -> ImageRequest:
"""
Uploads an image to Bing's AI service and returns the image response.
@@ -38,22 +35,22 @@ async def upload_image(
RuntimeError: If the image upload fails.
Returns:
- ImageResponse: The response from the image upload.
+ ImageRequest: The response from the image upload.
"""
image = to_image(image_data)
new_width, new_height = calculate_new_dimensions(image)
- processed_img = process_image(image, new_width, new_height)
- img_binary_data = to_base64(processed_img, IMAGE_CONFIG['imageCompressionRate'])
+ image = process_image(image, new_width, new_height)
+ img_binary_data = to_base64_jpg(image, IMAGE_CONFIG['imageCompressionRate'])
- data, boundary = build_image_upload_payload(img_binary_data, tone)
- headers = prepare_headers(session, boundary)
+ data = build_image_upload_payload(img_binary_data, tone)
+ headers = prepare_headers(session)
async with session.post("https://www.bing.com/images/kblob", data=data, headers=headers, proxy=proxy) as response:
if response.status != 200:
raise RuntimeError("Failed to upload image.")
return parse_image_response(await response.json())
-def calculate_new_dimensions(image: Image.Image) -> Tuple[int, int]:
+def calculate_new_dimensions(image: Image) -> Tuple[int, int]:
"""
Calculates the new dimensions for the image based on the maximum allowed pixels.
@@ -70,7 +67,7 @@ def calculate_new_dimensions(image: Image.Image) -> Tuple[int, int]:
return int(width * scale_factor), int(height * scale_factor)
return width, height
-def build_image_upload_payload(image_bin: str, tone: str) -> Tuple[str, str]:
+def build_image_upload_payload(image_bin: str, tone: str) -> FormData:
"""
Builds the payload for image uploading.
@@ -81,18 +78,11 @@ def build_image_upload_payload(image_bin: str, tone: str) -> Tuple[str, str]:
Returns:
Tuple[str, str]: The data and boundary for the payload.
"""
- boundary = "----WebKitFormBoundary" + ''.join(random.choices(string.ascii_letters + string.digits, k=16))
- data = f"""--{boundary}
-Content-Disposition: form-data; name="knowledgeRequest"
-
-{json.dumps(build_knowledge_request(tone), ensure_ascii=False)}
---{boundary}
-Content-Disposition: form-data; name="imageBase64"
-
-{image_bin}
---{boundary}--
-"""
- return data, boundary
+ data = FormData()
+ knowledge_request = json.dumps(build_knowledge_request(tone), ensure_ascii=False)
+ data.add_field('knowledgeRequest', knowledge_request, content_type="application/json")
+ data.add_field('imageBase64', image_bin)
+ return data
def build_knowledge_request(tone: str) -> dict:
"""
@@ -119,7 +109,7 @@ def build_knowledge_request(tone: str) -> dict:
}
}
-def prepare_headers(session: ClientSession, boundary: str) -> dict:
+def prepare_headers(session: ClientSession) -> dict:
"""
Prepares the headers for the image upload request.
@@ -131,12 +121,11 @@ def prepare_headers(session: ClientSession, boundary: str) -> dict:
dict: The headers for the request.
"""
headers = session.headers.copy()
- headers["Content-Type"] = f'multipart/form-data; boundary={boundary}'
headers["Referer"] = 'https://www.bing.com/search?q=Bing+AI&showconv=1&FORM=hpcodx'
headers["Origin"] = 'https://www.bing.com'
return headers
-def parse_image_response(response: dict) -> ImageResponse:
+def parse_image_response(response: dict) -> ImageRequest:
"""
Parses the response from the image upload.
@@ -147,7 +136,7 @@ def parse_image_response(response: dict) -> ImageResponse:
RuntimeError: If parsing the image info fails.
Returns:
- ImageResponse: The parsed image response.
+ ImageRequest: The parsed image response.
"""
if not response.get('blobId'):
raise RuntimeError("Failed to parse image info.")
@@ -160,4 +149,4 @@ def parse_image_response(response: dict) -> ImageResponse:
if IMAGE_CONFIG["enableFaceBlurDebug"] else
f"https://www.bing.com/images/blob?bcid={result['bcid']}"
)
- return ImageResponse(result["imageUrl"], "", result) \ No newline at end of file
+ return ImageRequest(result["imageUrl"], "", result) \ No newline at end of file
diff --git a/g4f/Provider/deprecated/ChatgptDuo.py b/g4f/Provider/deprecated/ChatgptDuo.py
index c2d2de7a..bd9e195d 100644
--- a/g4f/Provider/deprecated/ChatgptDuo.py
+++ b/g4f/Provider/deprecated/ChatgptDuo.py
@@ -1,7 +1,7 @@
from __future__ import annotations
from ...typing import Messages
-from curl_cffi.requests import AsyncSession
+from ...requests import StreamSession
from ..base_provider import AsyncProvider, format_prompt
@@ -19,7 +19,7 @@ class ChatgptDuo(AsyncProvider):
timeout: int = 120,
**kwargs
) -> str:
- async with AsyncSession(
+ async with StreamSession(
impersonate="chrome107",
proxies={"https": proxy},
timeout=timeout
diff --git a/g4f/Provider/deprecated/GetGpt.py b/g4f/Provider/deprecated/GetGpt.py
index 69851ee5..dd586569 100644
--- a/g4f/Provider/deprecated/GetGpt.py
+++ b/g4f/Provider/deprecated/GetGpt.py
@@ -5,10 +5,10 @@ import os
import uuid
import requests
-try:
- from Crypto.Cipher import AES
-except ImportError:
- from Cryptodome.Cipher import AES
+# try:
+# from Crypto.Cipher import AES
+# except ImportError:
+# from Cryptodome.Cipher import AES
from ...typing import Any, CreateResult
from ..base_provider import AbstractProvider
@@ -57,19 +57,21 @@ class GetGpt(AbstractProvider):
def _encrypt(e: str):
- t = os.urandom(8).hex().encode('utf-8')
- n = os.urandom(8).hex().encode('utf-8')
- r = e.encode('utf-8')
+ # t = os.urandom(8).hex().encode('utf-8')
+ # n = os.urandom(8).hex().encode('utf-8')
+ # r = e.encode('utf-8')
- cipher = AES.new(t, AES.MODE_CBC, n)
- ciphertext = cipher.encrypt(_pad_data(r))
+ # cipher = AES.new(t, AES.MODE_CBC, n)
+ # ciphertext = cipher.encrypt(_pad_data(r))
- return ciphertext.hex() + t.decode('utf-8') + n.decode('utf-8')
+ # return ciphertext.hex() + t.decode('utf-8') + n.decode('utf-8')
+ return
def _pad_data(data: bytes) -> bytes:
- block_size = AES.block_size
- padding_size = block_size - len(data) % block_size
- padding = bytes([padding_size] * padding_size)
+ # block_size = AES.block_size
+ # padding_size = block_size - len(data) % block_size
+ # padding = bytes([padding_size] * padding_size)
- return data + padding
+ # return data + padding
+ return \ No newline at end of file
diff --git a/g4f/Provider/helper.py b/g4f/Provider/helper.py
index cf204e39..0af61d8d 100644
--- a/g4f/Provider/helper.py
+++ b/g4f/Provider/helper.py
@@ -1,57 +1,37 @@
from __future__ import annotations
-import asyncio
import os
import random
import secrets
import string
-from asyncio import AbstractEventLoop, BaseEventLoop
from aiohttp import BaseConnector
-from platformdirs import user_config_dir
-from browser_cookie3 import (
- chrome, chromium, opera, opera_gx,
- brave, edge, vivaldi, firefox,
- _LinuxPasswordManager, BrowserCookieError
-)
+
+try:
+ from platformdirs import user_config_dir
+ has_platformdirs = True
+except ImportError:
+ has_platformdirs = False
+try:
+ from browser_cookie3 import (
+ chrome, chromium, opera, opera_gx,
+ brave, edge, vivaldi, firefox,
+ _LinuxPasswordManager, BrowserCookieError
+ )
+ has_browser_cookie3 = True
+except ImportError:
+ has_browser_cookie3 = False
+
from ..typing import Dict, Messages, Optional
-from ..errors import AiohttpSocksError
+from ..errors import AiohttpSocksError, MissingRequirementsError
from .. import debug
# Global variable to store cookies
_cookies: Dict[str, Dict[str, str]] = {}
-def get_event_loop() -> AbstractEventLoop:
- """
- Get the current asyncio event loop. If the loop is closed or not set, create a new event loop.
- If a loop is running, handle nested event loops. Patch the loop if 'nest_asyncio' is installed.
-
- Returns:
- AbstractEventLoop: The current or new event loop.
- """
- try:
- loop = asyncio.get_event_loop()
- if isinstance(loop, BaseEventLoop):
- loop._check_closed()
- except RuntimeError:
- loop = asyncio.new_event_loop()
- asyncio.set_event_loop(loop)
- try:
- asyncio.get_running_loop()
- if not hasattr(loop.__class__, "_nest_patched"):
- import nest_asyncio
- nest_asyncio.apply(loop)
- except RuntimeError:
- pass
- except ImportError:
- raise RuntimeError(
- 'Use "create_async" instead of "create" function in a running event loop. Or install "nest_asyncio" package.'
- )
- return loop
-
-if os.environ.get('DBUS_SESSION_BUS_ADDRESS') == "/dev/null":
+if has_browser_cookie3 and os.environ.get('DBUS_SESSION_BUS_ADDRESS') == "/dev/null":
_LinuxPasswordManager.get_password = lambda a, b: b"secret"
-def get_cookies(domain_name: str = '') -> Dict[str, str]:
+def get_cookies(domain_name: str = '', raise_requirements_error: bool = True) -> Dict[str, str]:
"""
Load cookies for a given domain from all supported browsers and cache the results.
@@ -64,11 +44,11 @@ def get_cookies(domain_name: str = '') -> Dict[str, str]:
if domain_name in _cookies:
return _cookies[domain_name]
- cookies = _load_cookies_from_browsers(domain_name)
+ cookies = load_cookies_from_browsers(domain_name, raise_requirements_error)
_cookies[domain_name] = cookies
return cookies
-def _load_cookies_from_browsers(domain_name: str) -> Dict[str, str]:
+def load_cookies_from_browsers(domain_name: str, raise_requirements_error: bool = True) -> Dict[str, str]:
"""
Helper function to load cookies from various browsers.
@@ -78,6 +58,10 @@ def _load_cookies_from_browsers(domain_name: str) -> Dict[str, str]:
Returns:
Dict[str, str]: A dictionary of cookie names and values.
"""
+ if not has_browser_cookie3:
+ if raise_requirements_error:
+ raise MissingRequirementsError('Install "browser_cookie3" package')
+ return {}
cookies = {}
for cookie_fn in [_g4f, chrome, chromium, opera, opera_gx, brave, edge, vivaldi, firefox]:
try:
@@ -104,6 +88,8 @@ def _g4f(domain_name: str) -> list:
Returns:
list: List of cookies.
"""
+ if not has_platformdirs:
+ return []
user_data_dir = user_config_dir("g4f")
cookie_file = os.path.join(user_data_dir, "Default", "Cookies")
return [] if not os.path.exists(cookie_file) else chrome(cookie_file, domain_name)
diff --git a/g4f/Provider/needs_auth/Bard.py b/g4f/Provider/needs_auth/Bard.py
index aea67874..09ed1c3c 100644
--- a/g4f/Provider/needs_auth/Bard.py
+++ b/g4f/Provider/needs_auth/Bard.py
@@ -2,10 +2,14 @@ from __future__ import annotations
import time
import os
-from selenium.webdriver.common.by import By
-from selenium.webdriver.support.ui import WebDriverWait
-from selenium.webdriver.support import expected_conditions as EC
-from selenium.webdriver.common.keys import Keys
+
+try:
+ from selenium.webdriver.common.by import By
+ from selenium.webdriver.support.ui import WebDriverWait
+ from selenium.webdriver.support import expected_conditions as EC
+ from selenium.webdriver.common.keys import Keys
+except ImportError:
+ pass
from ...typing import CreateResult, Messages
from ..base_provider import AbstractProvider
diff --git a/g4f/Provider/needs_auth/OpenaiChat.py b/g4f/Provider/needs_auth/OpenaiChat.py
index 85866272..b07bd49b 100644
--- a/g4f/Provider/needs_auth/OpenaiChat.py
+++ b/g4f/Provider/needs_auth/OpenaiChat.py
@@ -1,21 +1,32 @@
from __future__ import annotations
+
import asyncio
import uuid
import json
import os
-from py_arkose_generator.arkose import get_values_for_request
-from async_property import async_cached_property
-from selenium.webdriver.common.by import By
-from selenium.webdriver.support.ui import WebDriverWait
-from selenium.webdriver.support import expected_conditions as EC
+try:
+ from py_arkose_generator.arkose import get_values_for_request
+ from async_property import async_cached_property
+ has_requirements = True
+except ImportError:
+ async_cached_property = property
+ has_requirements = False
+try:
+ from selenium.webdriver.common.by import By
+ from selenium.webdriver.support.ui import WebDriverWait
+ from selenium.webdriver.support import expected_conditions as EC
+ has_webdriver = True
+except ImportError:
+ has_webdriver = False
from ..base_provider import AsyncGeneratorProvider, ProviderModelMixin
from ..helper import format_prompt, get_cookies
from ...webdriver import get_browser, get_driver_cookies
-from ...typing import AsyncResult, Messages
+from ...typing import AsyncResult, Messages, Cookies, ImageType
from ...requests import StreamSession
-from ...image import to_image, to_bytes, ImageType, ImageResponse
+from ...image import to_image, to_bytes, ImageResponse, ImageRequest
+from ...errors import MissingRequirementsError, MissingAccessToken
class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
@@ -27,12 +38,8 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
supports_gpt_35_turbo = True
supports_gpt_4 = True
default_model = None
- models = ["text-davinci-002-render-sha", "gpt-4", "gpt-4-gizmo"]
- model_aliases = {
- "gpt-3.5-turbo": "text-davinci-002-render-sha",
- }
+ models = ["gpt-3.5-turbo", "gpt-4", "gpt-4-gizmo"]
_cookies: dict = {}
- _default_model: str = None
@classmethod
async def create(
@@ -94,7 +101,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
session: StreamSession,
headers: dict,
image: ImageType
- ) -> ImageResponse:
+ ) -> ImageRequest:
"""
Upload an image to the service and get the download URL
@@ -104,7 +111,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
image: The image to upload, either a PIL Image object or a bytes object
Returns:
- An ImageResponse object that contains the download URL, file name, and other data
+ An ImageRequest object that contains the download URL, file name, and other data
"""
# Convert the image to a PIL Image object and get the extension
image = to_image(image)
@@ -145,7 +152,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
) as response:
response.raise_for_status()
download_url = (await response.json())["download_url"]
- return ImageResponse(download_url, image_data["file_name"], image_data)
+ return ImageRequest(download_url, image_data["file_name"], image_data)
@classmethod
async def get_default_model(cls, session: StreamSession, headers: dict):
@@ -169,7 +176,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
return cls.default_model
@classmethod
- def create_messages(cls, prompt: str, image_response: ImageResponse = None):
+ def create_messages(cls, prompt: str, image_response: ImageRequest = None):
"""
Create a list of messages for the user input
@@ -282,7 +289,7 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
proxy: str = None,
timeout: int = 120,
access_token: str = None,
- cookies: dict = None,
+ cookies: Cookies = None,
auto_continue: bool = False,
history_disabled: bool = True,
action: str = "next",
@@ -317,12 +324,16 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
Raises:
RuntimeError: If an error occurs during processing.
"""
+ if not has_requirements:
+ raise MissingRequirementsError('Install "py-arkose-generator" and "async_property" package')
if not parent_id:
parent_id = str(uuid.uuid4())
if not cookies:
- cookies = cls._cookies or get_cookies("chat.openai.com")
+ cookies = cls._cookies or get_cookies("chat.openai.com", False)
if not access_token and "access_token" in cookies:
access_token = cookies["access_token"]
+ if not access_token and not has_webdriver:
+ raise MissingAccessToken(f'Missing "access_token"')
if not access_token:
login_url = os.environ.get("G4F_LOGIN_URL")
if login_url:
@@ -331,7 +342,6 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
cls._cookies = cookies
headers = {"Authorization": f"Bearer {access_token}"}
-
async with StreamSession(
proxies={"https": proxy},
impersonate="chrome110",
@@ -346,13 +356,15 @@ class OpenaiChat(AsyncGeneratorProvider, ProviderModelMixin):
except Exception as e:
yield e
end_turn = EndTurn()
+ model = cls.get_model(model or await cls.get_default_model(session, headers))
+ model = "text-davinci-002-render-sha" if model == "gpt-3.5-turbo" else model
while not end_turn.is_end:
data = {
"action": action,
"arkose_token": await cls.get_arkose_token(session),
"conversation_id": conversation_id,
"parent_message_id": parent_id,
- "model": cls.get_model(model or await cls.get_default_model(session, headers)),
+ "model": model,
"history_and_training_disabled": history_disabled and not auto_continue,
}
if action != "continue":
diff --git a/g4f/Provider/selenium/PerplexityAi.py b/g4f/Provider/selenium/PerplexityAi.py
index 4796f709..8ae6ad2b 100644
--- a/g4f/Provider/selenium/PerplexityAi.py
+++ b/g4f/Provider/selenium/PerplexityAi.py
@@ -1,10 +1,14 @@
from __future__ import annotations
import time
-from selenium.webdriver.common.by import By
-from selenium.webdriver.support.ui import WebDriverWait
-from selenium.webdriver.support import expected_conditions as EC
-from selenium.webdriver.common.keys import Keys
+
+try:
+ from selenium.webdriver.common.by import By
+ from selenium.webdriver.support.ui import WebDriverWait
+ from selenium.webdriver.support import expected_conditions as EC
+ from selenium.webdriver.common.keys import Keys
+except ImportError:
+ pass
from ...typing import CreateResult, Messages
from ..base_provider import AbstractProvider
diff --git a/g4f/base_provider.py b/g4f/base_provider.py
index 03ae64d6..cc3451a2 100644
--- a/g4f/base_provider.py
+++ b/g4f/base_provider.py
@@ -1,3 +1,5 @@
+from __future__ import annotations
+
from abc import ABC, abstractmethod
from typing import Union, List, Dict, Type
from .typing import Messages, CreateResult
diff --git a/g4f/cli.py b/g4f/cli.py
index 40ddd3d7..64d63fd3 100644
--- a/g4f/cli.py
+++ b/g4f/cli.py
@@ -4,7 +4,6 @@ from enum import Enum
import g4f
from g4f import Provider
-from g4f.api import Api
from g4f.gui.run import gui_parser, run_gui_args
def run_gui(args):
@@ -23,6 +22,7 @@ def main():
args = parser.parse_args()
if args.mode == "api":
+ from g4f.api import Api
controller=Api(engine=g4f, debug=args.debug, list_ignored_providers=args.ignored_providers)
controller.run(args.bind)
elif args.mode == "gui":
diff --git a/g4f/errors.py b/g4f/errors.py
index aaaa6b4b..66d41d4c 100644
--- a/g4f/errors.py
+++ b/g4f/errors.py
@@ -31,5 +31,11 @@ class NestAsyncioError(Exception):
class ModelNotSupportedError(Exception):
pass
-class AiohttpSocksError(Exception):
+class MissingRequirementsError(Exception):
+ pass
+
+class AiohttpSocksError(MissingRequirementsError):
+ pass
+
+class MissingAccessToken(Exception):
pass \ No newline at end of file
diff --git a/g4f/gui/__init__.py b/g4f/gui/__init__.py
index a8000e71..d2dfbbc5 100644
--- a/g4f/gui/__init__.py
+++ b/g4f/gui/__init__.py
@@ -1,6 +1,10 @@
-from .server.app import app
-from .server.website import Website
-from .server.backend import Backend_Api
+try:
+ from .server.app import app
+ from .server.website import Website
+ from .server.backend import Backend_Api
+except ImportError:
+ from g4f.errors import MissingRequirementsError
+ raise MissingRequirementsError('Install "flask" and "werkzeug" package for gui')
def run_gui(host: str = '0.0.0.0', port: int = 80, debug: bool = False) -> None:
config = {
diff --git a/g4f/gui/run.py b/g4f/gui/run.py
index 7ff769fd..6b7d423f 100644
--- a/g4f/gui/run.py
+++ b/g4f/gui/run.py
@@ -1,8 +1,5 @@
from argparse import ArgumentParser
-from g4f.gui import run_gui
-
-
def gui_parser():
parser = ArgumentParser(description="Run the GUI")
parser.add_argument("-host", type=str, default="0.0.0.0", help="hostname")
@@ -10,15 +7,14 @@ def gui_parser():
parser.add_argument("-debug", action="store_true", help="debug mode")
return parser
-
def run_gui_args(args):
+ from g4f.gui import run_gui
host = args.host
port = args.port
debug = args.debug
run_gui(host, port, debug)
-
if __name__ == "__main__":
parser = gui_parser()
args = parser.parse_args()
- run_gui_args(args)
+ run_gui_args(args) \ No newline at end of file
diff --git a/g4f/gui/server/internet.py b/g4f/gui/server/internet.py
index 6c2e3a89..a6bfc885 100644
--- a/g4f/gui/server/internet.py
+++ b/g4f/gui/server/internet.py
@@ -1,8 +1,14 @@
from __future__ import annotations
-from bs4 import BeautifulSoup
from aiohttp import ClientSession, ClientTimeout
-from duckduckgo_search import DDGS
+try:
+ from duckduckgo_search import DDGS
+ from bs4 import BeautifulSoup
+ has_requirements = True
+except ImportError:
+ has_requirements = False
+from ...errors import MissingRequirementsError
+
import asyncio
class SearchResults():
@@ -88,6 +94,8 @@ async def fetch_and_scrape(session: ClientSession, url: str, max_words: int = No
return
async def search(query: str, n_results: int = 5, max_words: int = 2500, add_text: bool = True) -> SearchResults:
+ if not has_requirements:
+ raise MissingRequirementsError('Install "duckduckgo-search" and "beautifulsoup4" package')
with DDGS() as ddgs:
results = []
for result in ddgs.text(
diff --git a/g4f/image.py b/g4f/image.py
index 94b8c24c..68767155 100644
--- a/g4f/image.py
+++ b/g4f/image.py
@@ -1,39 +1,52 @@
+from __future__ import annotations
+
import re
from io import BytesIO
import base64
from .typing import ImageType, Union
-from PIL import Image
+
+try:
+ from PIL.Image import open as open_image, new as new_image, Image
+ from PIL.Image import FLIP_LEFT_RIGHT, ROTATE_180, ROTATE_270, ROTATE_90
+ has_requirements = True
+except ImportError:
+ Image = type
+ has_requirements = False
+
+from .errors import MissingRequirementsError
ALLOWED_EXTENSIONS = {'png', 'jpg', 'jpeg', 'gif', 'webp', 'svg'}
-def to_image(image: ImageType, is_svg: bool = False) -> Image.Image:
+def to_image(image: ImageType, is_svg: bool = False) -> Image:
"""
Converts the input image to a PIL Image object.
Args:
- image (Union[str, bytes, Image.Image]): The input image.
+ image (Union[str, bytes, Image]): The input image.
Returns:
- Image.Image: The converted PIL Image object.
+ Image: The converted PIL Image object.
"""
+ if not has_requirements:
+ raise MissingRequirementsError('Install "pillow" package for images')
if is_svg:
try:
import cairosvg
except ImportError:
- raise RuntimeError('Install "cairosvg" package for svg images')
+ raise MissingRequirementsError('Install "cairosvg" package for svg images')
if not isinstance(image, bytes):
image = image.read()
buffer = BytesIO()
cairosvg.svg2png(image, write_to=buffer)
- return Image.open(buffer)
+ return open_image(buffer)
if isinstance(image, str):
is_data_uri_an_image(image)
image = extract_data_uri(image)
if isinstance(image, bytes):
is_accepted_format(image)
- return Image.open(BytesIO(image))
- elif not isinstance(image, Image.Image):
- image = Image.open(image)
+ return open_image(BytesIO(image))
+ elif not isinstance(image, Image):
+ image = open_image(image)
copy = image.copy()
copy.format = image.format
return copy
@@ -110,12 +123,12 @@ def extract_data_uri(data_uri: str) -> bytes:
data = base64.b64decode(data)
return data
-def get_orientation(image: Image.Image) -> int:
+def get_orientation(image: Image) -> int:
"""
Gets the orientation of the given image.
Args:
- image (Image.Image): The image.
+ image (Image): The image.
Returns:
int: The orientation value.
@@ -126,40 +139,40 @@ def get_orientation(image: Image.Image) -> int:
if orientation is not None:
return orientation
-def process_image(img: Image.Image, new_width: int, new_height: int) -> Image.Image:
+def process_image(img: Image, new_width: int, new_height: int) -> Image:
"""
Processes the given image by adjusting its orientation and resizing it.
Args:
- img (Image.Image): The image to process.
+ img (Image): The image to process.
new_width (int): The new width of the image.
new_height (int): The new height of the image.
Returns:
- Image.Image: The processed image.
+ Image: The processed image.
"""
# Fix orientation
orientation = get_orientation(img)
if orientation:
if orientation > 4:
- img = img.transpose(Image.FLIP_LEFT_RIGHT)
+ img = img.transpose(FLIP_LEFT_RIGHT)
if orientation in [3, 4]:
- img = img.transpose(Image.ROTATE_180)
+ img = img.transpose(ROTATE_180)
if orientation in [5, 6]:
- img = img.transpose(Image.ROTATE_270)
+ img = img.transpose(ROTATE_270)
if orientation in [7, 8]:
- img = img.transpose(Image.ROTATE_90)
+ img = img.transpose(ROTATE_90)
# Resize image
img.thumbnail((new_width, new_height))
# Remove transparency
if img.mode != "RGB":
img.load()
- white = Image.new('RGB', img.size, (255, 255, 255))
+ white = new_image('RGB', img.size, (255, 255, 255))
white.paste(img, mask=img.split()[3])
return white
return img
-def to_base64(image: Image.Image, compression_rate: float) -> str:
+def to_base64_jpg(image: Image, compression_rate: float) -> str:
"""
Converts the given image to a base64-encoded string.
@@ -195,7 +208,7 @@ def format_images_markdown(images, alt: str, preview: str="{image}?w=200&h=200")
end_flag = "<!-- generated images end -->\n"
return f"\n{start_flag}{images}\n{end_flag}\n"
-def to_bytes(image: Image.Image) -> bytes:
+def to_bytes(image: Image) -> bytes:
"""
Converts the given image to bytes.
@@ -225,4 +238,7 @@ class ImageResponse():
return format_images_markdown(self.images, self.alt)
def get(self, key: str):
- return self.options.get(key) \ No newline at end of file
+ return self.options.get(key)
+
+class ImageRequest(ImageResponse):
+ pass \ No newline at end of file
diff --git a/g4f/requests.py b/g4f/requests.py
index 466d5a2a..46f3b457 100644
--- a/g4f/requests.py
+++ b/g4f/requests.py
@@ -4,80 +4,124 @@ import json
from functools import partialmethod
from typing import AsyncGenerator
from urllib.parse import urlparse
-from curl_cffi.requests import AsyncSession, Session, Response
-from .webdriver import WebDriver, WebDriverSession, bypass_cloudflare, get_driver_cookies
-
-class StreamResponse:
- """
- A wrapper class for handling asynchronous streaming responses.
-
- Attributes:
- inner (Response): The original Response object.
- """
- def __init__(self, inner: Response) -> None:
- """Initialize the StreamResponse with the provided Response object."""
- self.inner: Response = inner
-
- async def text(self) -> str:
- """Asynchronously get the response text."""
- return await self.inner.atext()
-
- def raise_for_status(self) -> None:
- """Raise an HTTPError if one occurred."""
- self.inner.raise_for_status()
-
- async def json(self, **kwargs) -> dict:
- """Asynchronously parse the JSON response content."""
- return json.loads(await self.inner.acontent(), **kwargs)
-
- async def iter_lines(self) -> AsyncGenerator[bytes, None]:
- """Asynchronously iterate over the lines of the response."""
- async for line in self.inner.aiter_lines():
- yield line
-
- async def iter_content(self) -> AsyncGenerator[bytes, None]:
- """Asynchronously iterate over the response content."""
- async for chunk in self.inner.aiter_content():
- yield chunk
-
- async def __aenter__(self):
- """Asynchronously enter the runtime context for the response object."""
- inner: Response = await self.inner
- self.inner = inner
- self.request = inner.request
- self.status_code: int = inner.status_code
- self.reason: str = inner.reason
- self.ok: bool = inner.ok
- self.headers = inner.headers
- self.cookies = inner.cookies
- return self
-
- async def __aexit__(self, *args):
- """Asynchronously exit the runtime context for the response object."""
- await self.inner.aclose()
-
-
-class StreamSession(AsyncSession):
- """
- An asynchronous session class for handling HTTP requests with streaming.
-
- Inherits from AsyncSession.
- """
+try:
+ from curl_cffi.requests import AsyncSession, Session, Response
+ has_curl_cffi = True
+except ImportError:
+ Session = type
+ has_curl_cffi = False
- def request(
- self, method: str, url: str, **kwargs
- ) -> StreamResponse:
- """Create and return a StreamResponse object for the given HTTP request."""
- return StreamResponse(super().request(method, url, stream=True, **kwargs))
-
- # Defining HTTP methods as partial methods of the request method.
- head = partialmethod(request, "HEAD")
- get = partialmethod(request, "GET")
- post = partialmethod(request, "POST")
- put = partialmethod(request, "PUT")
- patch = partialmethod(request, "PATCH")
- delete = partialmethod(request, "DELETE")
+from .webdriver import WebDriver, WebDriverSession, bypass_cloudflare, get_driver_cookies
+from .errors import MissingRequirementsError
+
+if not has_curl_cffi:
+ from aiohttp import ClientSession, ClientResponse, ClientTimeout
+ from .Provider.helper import get_connector
+
+ class StreamResponse(ClientResponse):
+ async def iter_lines(self) -> iter[bytes, None]:
+ async for line in self.content:
+ yield line.rstrip(b"\r\n")
+
+ async def json(self):
+ return await super().json(content_type=None)
+
+ class StreamSession(ClientSession):
+ def __init__(self, headers: dict = {}, timeout: int = None, proxies: dict = {}, impersonate = None, **kwargs):
+ if impersonate:
+ headers = {
+ 'Accept-Encoding': 'gzip, deflate, br',
+ 'Accept-Language': 'en-US',
+ 'Connection': 'keep-alive',
+ 'Sec-Fetch-Dest': 'empty',
+ 'Sec-Fetch-Mode': 'cors',
+ 'Sec-Fetch-Site': 'same-site',
+ "User-Agent": 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/107.0.0.0 Safari/537.36',
+ 'Accept': '*/*',
+ 'sec-ch-ua': '"Google Chrome";v="107", "Chromium";v="107", "Not?A_Brand";v="24"',
+ 'sec-ch-ua-mobile': '?0',
+ 'sec-ch-ua-platform': '"Windows"',
+ **headers
+ }
+ super().__init__(
+ **kwargs,
+ timeout=ClientTimeout(timeout) if timeout else None,
+ response_class=StreamResponse,
+ connector=get_connector(kwargs.get("connector"), proxies.get("https")),
+ headers=headers
+ )
+else:
+ class StreamResponse:
+ """
+ A wrapper class for handling asynchronous streaming responses.
+
+ Attributes:
+ inner (Response): The original Response object.
+ """
+
+ def __init__(self, inner: Response) -> None:
+ """Initialize the StreamResponse with the provided Response object."""
+ self.inner: Response = inner
+
+ async def text(self) -> str:
+ """Asynchronously get the response text."""
+ return await self.inner.atext()
+
+ def raise_for_status(self) -> None:
+ """Raise an HTTPError if one occurred."""
+ self.inner.raise_for_status()
+
+ async def json(self, **kwargs) -> dict:
+ """Asynchronously parse the JSON response content."""
+ return json.loads(await self.inner.acontent(), **kwargs)
+
+ async def iter_lines(self) -> AsyncGenerator[bytes, None]:
+ """Asynchronously iterate over the lines of the response."""
+ async for line in self.inner.aiter_lines():
+ yield line
+
+ async def iter_content(self) -> AsyncGenerator[bytes, None]:
+ """Asynchronously iterate over the response content."""
+ async for chunk in self.inner.aiter_content():
+ yield chunk
+
+ async def __aenter__(self):
+ """Asynchronously enter the runtime context for the response object."""
+ inner: Response = await self.inner
+ self.inner = inner
+ self.request = inner.request
+ self.status_code: int = inner.status_code
+ self.reason: str = inner.reason
+ self.ok: bool = inner.ok
+ self.headers = inner.headers
+ self.cookies = inner.cookies
+ return self
+
+ async def __aexit__(self, *args):
+ """Asynchronously exit the runtime context for the response object."""
+ await self.inner.aclose()
+
+ class StreamSession(AsyncSession):
+ """
+ An asynchronous session class for handling HTTP requests with streaming.
+
+ Inherits from AsyncSession.
+ """
+
+ def request(
+ self, method: str, url: str, **kwargs
+ ) -> StreamResponse:
+ """Create and return a StreamResponse object for the given HTTP request."""
+ return StreamResponse(super().request(method, url, stream=True, **kwargs))
+
+ # Defining HTTP methods as partial methods of the request method.
+ head = partialmethod(request, "HEAD")
+ get = partialmethod(request, "GET")
+ post = partialmethod(request, "POST")
+ put = partialmethod(request, "PUT")
+ patch = partialmethod(request, "PATCH")
+ delete = partialmethod(request, "DELETE")
def get_session_from_browser(url: str, webdriver: WebDriver = None, proxy: str = None, timeout: int = 120) -> Session:
@@ -93,6 +137,8 @@ def get_session_from_browser(url: str, webdriver: WebDriver = None, proxy: str =
Returns:
Session: A Session object configured with cookies and headers from the WebDriver.
"""
+ if not has_curl_cffi:
+ raise MissingRequirementsError('Install "curl_cffi" package')
with WebDriverSession(webdriver, "", proxy=proxy, virtual_display=True) as driver:
bypass_cloudflare(driver, url, timeout)
cookies = get_driver_cookies(driver)
diff --git a/g4f/typing.py b/g4f/typing.py
index fd4fd047..c5a981bd 100644
--- a/g4f/typing.py
+++ b/g4f/typing.py
@@ -1,6 +1,9 @@
import sys
from typing import Any, AsyncGenerator, Generator, NewType, Tuple, Union, List, Dict, Type, IO, Optional
-from PIL.Image import Image
+try:
+ from PIL.Image import Image
+except ImportError:
+ Image = type
if sys.version_info >= (3, 8):
from typing import TypedDict
@@ -11,6 +14,7 @@ SHA256 = NewType('sha_256_hash', str)
CreateResult = Generator[str, None, None]
AsyncResult = AsyncGenerator[str, None]
Messages = List[Dict[str, str]]
+Cookies = List[Dict[str, str]]
ImageType = Union[str, bytes, IO, Image, None]
__all__ = [
diff --git a/g4f/version.py b/g4f/version.py
index 9201c75c..63941baf 100644
--- a/g4f/version.py
+++ b/g4f/version.py
@@ -1,3 +1,5 @@
+from __future__ import annotations
+
from os import environ
import requests
from functools import cached_property
diff --git a/g4f/webdriver.py b/g4f/webdriver.py
index 85d6d695..d2009ddc 100644
--- a/g4f/webdriver.py
+++ b/g4f/webdriver.py
@@ -1,12 +1,20 @@
from __future__ import annotations
-from platformdirs import user_config_dir
-from selenium.webdriver.remote.webdriver import WebDriver
-from undetected_chromedriver import Chrome, ChromeOptions
-from selenium.webdriver.common.by import By
-from selenium.webdriver.support.ui import WebDriverWait
-from selenium.webdriver.support import expected_conditions as EC
+
+try:
+ from platformdirs import user_config_dir
+ from selenium.webdriver.remote.webdriver import WebDriver
+ from undetected_chromedriver import Chrome, ChromeOptions
+ from selenium.webdriver.common.by import By
+ from selenium.webdriver.support.ui import WebDriverWait
+ from selenium.webdriver.support import expected_conditions as EC
+ has_requirements = True
+except ImportError:
+ WebDriver = type
+ has_requirements = False
+
from os import path
from os import access, R_OK
+from .errors import MissingRequirementsError
from . import debug
try:
@@ -33,6 +41,8 @@ def get_browser(
Returns:
WebDriver: An instance of WebDriver configured with the specified options.
"""
+ if not has_requirements:
+ raise MissingRequirementsError('Install "undetected_chromedriver" and "platformdirs" package')
if user_data_dir is None:
user_data_dir = user_config_dir("g4f")
if user_data_dir and debug.logging:
@@ -144,7 +154,7 @@ class WebDriverSession:
Returns:
WebDriver: The reopened WebDriver instance.
"""
- user_data_dir = user_data_data_dir or self.user_data_dir
+ user_data_dir = user_data_dir or self.user_data_dir
if self.default_driver:
self.default_driver.quit()
if not virtual_display and self.virtual_display:
diff --git a/requirements-min.txt b/requirements-min.txt
new file mode 100644
index 00000000..a0402ccc
--- /dev/null
+++ b/requirements-min.txt
@@ -0,0 +1,2 @@
+requests
+aiohttp \ No newline at end of file
diff --git a/requirements.txt b/requirements.txt
index 12a7aec8..79bda967 100644
--- a/requirements.txt
+++ b/requirements.txt
@@ -4,7 +4,6 @@ curl_cffi>=0.5.10
aiohttp
certifi
browser_cookie3
-typing-extensions
PyExecJS
duckduckgo-search
nest_asyncio
@@ -16,7 +15,6 @@ fastapi
uvicorn
flask
py-arkose-generator
-asyncstdlib
async-property
undetected-chromedriver
brotli
diff --git a/setup.py b/setup.py
index 34f083ee..4a9c276f 100644
--- a/setup.py
+++ b/setup.py
@@ -8,33 +8,59 @@ here = os.path.abspath(os.path.dirname(__file__))
with codecs.open(os.path.join(here, 'README.md'), encoding='utf-8') as fh:
long_description = '\n' + fh.read()
-install_requires = [
+INSTALL_REQUIRE = [
"requests",
- "pycryptodome",
- "curl_cffi>=0.5.10",
"aiohttp",
- "certifi",
- "browser_cookie3",
- "typing-extensions",
- "PyExecJS",
- "duckduckgo-search",
- "nest_asyncio",
- "werkzeug",
- "loguru",
- "pillow",
- "platformdirs",
- "fastapi",
- "uvicorn",
- "flask",
- "py-arkose-generator",
- "asyncstdlib",
- "async-property",
- "undetected-chromedriver",
- "brotli",
- "beautifulsoup4",
- "setuptools",
]
+EXTRA_REQUIRE = {
+ 'all': [
+ "curl_cffi>=0.5.10",
+ "certifi",
+ "async-property", # openai
+ "py-arkose-generator", # openai
+ "browser_cookie3", # get_cookies
+ "PyExecJS", # GptForLove
+ "duckduckgo-search", # internet.search
+ "beautifulsoup4", # internet.search and bing.create_images
+ "brotli", # openai
+ "platformdirs", # webdriver
+ "undetected-chromedriver", # webdriver
+ "setuptools", # webdriver
+ "aiohttp_socks" # proxy
+ "pillow", # image
+ "cairosvg", # svg image
+ "werkzeug", "flask", # gui
+ "loguru", "fastapi",
+ "uvicorn", "nest_asyncio", # api
+ ],
+ "image": [
+ "pillow",
+ "cairosvg",
+ "beautifulsoup4"
+ ],
+ "webdriver": [
+ "platformdirs",
+ "undetected-chromedriver",
+ "setuptools"
+ ],
+ "openai": [
+ "async-property",
+ "py-arkose-generator",
+ "brotli"
+ ],
+ "api": [
+ "loguru", "fastapi",
+ "uvicorn", "nest_asyncio"
+ ],
+ "gui": [
+ "werkzeug", "flask",
+ "beautifulsoup4", "pillow",
+ "duckduckgo-search",
+ "browser_cookie3"
+ ]
+}
+
DESCRIPTION = (
'The official gpt4free repository | various collection of powerful language models'
)
@@ -53,7 +79,8 @@ setup(
'g4f': ['g4f/interference/*', 'g4f/gui/client/*', 'g4f/gui/server/*', 'g4f/Provider/npm/*']
},
include_package_data=True,
- install_requires=install_requires,
+ install_requires=INSTALL_REQUIRE,
+ extras_require=EXTRA_REQUIRE,
entry_points={
'console_scripts': ['g4f=g4f.cli:main'],
},